Learning Disentangled Attribute Representations for Robust Pedestrian Attribute Recognition

نویسندگان

چکیده

Although various methods have been proposed for pedestrian attribute recognition, most studies follow the same feature learning mechanism, \ie, a shared image to classify multiple attributes. However, this mechanism leads low-confidence predictions and non-robustness of model in inference stage. In paper, we investigate why is case. We mathematically discover that central cause optimal cannot maintain high similarities with classifiers simultaneously context minimizing classification loss. addition, ignores spatial semantic distinctions between different To address these limitations, propose novel disentangled (DAFL) framework learn each attribute, which exploits characteristics The mainly consists learnable queries, cascaded semantic-spatial cross-attention (SSCA) module, group attention merging (GAM) module. Specifically, based on SSCA module iteratively enhances localization attribute-related regions aggregates region features into features, used updating queries. GAM splits attributes groups distribution utilizes reliable supervise query maps. Experiments PETA, RAPv1, PA100k, RAPv2 show method performs favorably against state-of-the-art methods.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

DNA-GAN: Learning Disentangled Representations from Multi-Attribute Images

Disentangling factors of variation has always been a challenging problem in representation learning. Existing algorithms suffer from many limitations, such as unpredictable disentangling factors, bad quality of generated images from encodings, lack of identity information, etc. In this paper, we propose a supervised algorithm called DNA-GAN trying to disentangle different attributes of images. ...

متن کامل

Learning to Recognize Pedestrian Attribute

Learning to recognize pedestrian attributes at far distance is a novel research topic in video surveillance scenarios where face and body close-shots are hardly available; instead, only far-view video frames of pedestrian are given. In this study, we present an alternative approach that exploits the context of neighboring pedestrian images for improved attribute inference compared to the conven...

متن کامل

Sparse Representations and Distance Learning for Attribute Based Category Recognition

While traditional approaches in object recognition require the specification of training examples from each class and the application of class specific classifiers, in real world situations, the immensity of the number of image classes makes this task daunting. A novel approach in object recognition is attribute based classification, where instead of training classifiers for the recognition of ...

متن کامل

A Richly Annotated Dataset for Pedestrian Attribute Recognition

In this paper, we aim to improve the dataset foundation for pedestrian attribute recognition in real surveillance scenarios. Recognition of human attributes, such as gender, and clothes types, has great prospects in real applications. However, the development of suitable benchmark datasets for attribute recognition remains lagged behind. Existing human attribute datasets are collected from vari...

متن کامل

Learning Attribute Representation for Human Activity Recognition

Attribute representations became relevant in image recognition and word spotting, providing support under the presence of unbalance and disjoint datasets. However, for human activity recognition using sequential data from on-body sensors, human-labeled attributes are lacking. This paper introduces a search for attributes that represent favorably signal segments for recognizing human activities....

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2022

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v36i1.19991